Goto

Collaborating Authors

 random 0




LLossfunction b Batchsize ETotaltrainingepochs LEpochinterval SDA gfeatureextractor hpredictorfunction Composition Table4: Notations

Neural Information Processing Systems

Hence, the above formulation of set function is submodular and is an instance of concave over modular function. Inour setting, we use labeled target dataDt asthe validation set. The completed-SNE loss is defined as a combination of Land cross-entropy loss on source and targetdomain. Table 8 shows the training times for this setting. Again, we see that all instantiationsofORIENTachieve 2.5 speed-upcomparedtoFull.






LearningtoSeebyLookingatNoise-Supplementary Material

Neural Information Processing Systems

Dead leaves - Textures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Wetrain with stochastic gradient descent with momentum (set to0.9)for200epochs, starting with a learning rate of0.36 and decaying itby afactor of0.1atepochs155,170and185. The dimensionality of the last and the penultimate embedding are 128 and 4096 respectively. From left to right the columns correspond to the tasks: EuroSAT, Resisc45, Diabetic Retinopathy and Patch Camelyon. Here, wepresent additional data forthese experiments, and provide thefull distributions forthese criteria and all datasets.